Motif-based Hessian matrix forab initiogeometry optimization of nanostructures
نویسندگان
چکیده
منابع مشابه
Edge Detection with Hessian Matrix Property Based on Wavelet Transform
In this paper, we present an edge detection method based on wavelet transform and Hessian matrix of image at each pixel. Many methods which based on wavelet transform, use wavelet transform to approximate the gradient of image and detect edges by searching the modulus maximum of gradient vectors. In our scheme, we use wavelet transform to approximate Hessian matrix of image at each pixel, too. ...
متن کاملedge detection with hessian matrix property based on wavelet transform
in this paper, we present an edge detection method based on wavelet transform and hessian matrix of image at each pixel. many methods which based on wavelet transform, use wavelet transform to approximate the gradient of image and detect edges by searching the modulus maximum of gradient vectors. in our scheme, we use wavelet transform to approximate hessian matrix of image at each pixel, too. ...
متن کاملEar Structure Feature Extraction Based on Multi-scale Hessian Matrix
In this paper, a new ear anatomy feature edge extraction method based on Hessian matrix is proposed. Stable edge is obtained from principal curvature image across scale space. Firstly, the side face image that includes an ear is filtered and forms Gaussian pyramid. Secondly, the 2D gray image in the pyramid was regarded as a surface, maximum and minimum principal curvature and their direction w...
متن کاملPreconditioning for Hessian-Free Optimization
Recently Martens adapted the Hessian-free optimization method for the training of deep neural networks. One key aspect of this approach is that the Hessian is never computed explicitly, instead the Conjugate Gradient(CG) Algorithm is used to compute the new search direction by applying only matrix-vector products of the Hessian with arbitrary vectors. This can be done efficiently using a varian...
متن کاملBlock-diagonal Hessian-free Optimization
Second-order methods for neural network optimization have several advantages over methods based on first-order gradient descent, including better scaling to large mini-batch sizes and fewer updates needed for convergence. But they are rarely applied to deep learning in practice because of high computational cost and the need for model-dependent algorithmic variations. We introduce a variant of ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Physical Review B
سال: 2006
ISSN: 1098-0121,1550-235X
DOI: 10.1103/physrevb.73.193309